Stochastic and Adversarial Online Learning without Hyperparameters
نویسندگان
چکیده
Most online optimization algorithms focus on one of two things: performing well in adversarial settings by adapting to unknown data parameters (such as Lipschitz constants), typically achieving O( √ T ) regret, or performing well in stochastic settings where they can leverage some structure in the losses (such as strong convexity), typically achieving O(log(T )) regret. Algorithms that focus on the former problem hitherto achieved O( √ T ) in the stochastic setting rather than O(log(T )). Here we introduce an online optimization algorithm that achieves O(log(T )) regret in a wide class of stochastic settings while gracefully degrading to the optimal O( √ T ) regret in adversarial settings (up to logarithmic factors). Our algorithm does not require any prior knowledge about the data or tuning of parameters to achieve superior performance. 1 Extending Adversarial Algorithms to Stochastic Settings The online convex optimization (OCO) paradigm [1, 2] can be used to model a large number of scenarios of interest, such as streaming problems, adversarial environments, or stochastic optimization. In brief, an OCO algorithm plays T rounds of a game in which on each round the algorithm outputs a vector wt in some convex space W , and then receives a loss function `t :W → R that is convex. The algorithm’s objective is to minimize regret, which is the total loss of all rounds relative to w, the minimizer of ∑T t=1 `t in W : RT (w ) = T ∑
منابع مشابه
Owed to a Martingale: A Fast Bayesian On-Line EM Algorithm for Multinomial Models
This paper introduces a fast Bayesian online expectation maximization (BOEM) algorithm for multinomial mixtures. Using some properties of the Dirichlet distribution, we derive expressions for adaptive learning rates that depend solely on the data and the prior’s hyperparameters. As a result, we avoid the problem of having to tune the learning rates using heuristics. In the application to multin...
متن کاملHybrid Stochastic-Adversarial On-line Learning
Most of the research in online learning focused either on the problem of adversarial classification (i.e., both inputs and labels are arbitrarily chosen by an adversary) or on the traditional supervised learning problem in which samples are independently generated from a fixed probability distribution. Nonetheless, in a number of domains the relationship between inputs and labels may be adversa...
متن کاملUnderstanding the Energy and Precision Requirements for Online Learning
It is well-known that the precision of data, hyperparameters, and internal representations employed in learning systems directly impacts their energy, throughput, and latency. The precision requirements for the training algorithm are also important for systems that learn on-the-fly. Prior work has shown that the data and hyperparameters can be quantized heavily without incurring much penalty in...
متن کاملHot Swapping for Online Adaptation of Optimization Hyperparameters
We describe a general framework for online adaptation of optimization hyperparameters by ‘hot swapping’ their values during learning. We investigate this approach in the context of adaptive learning rate selection using an explore-exploit strategy from the multi-armed bandit literature. Experiments on a benchmark neural network show that the hot swapping approach leads to consistently better so...
متن کاملWithout-Replacement Sampling for Stochastic Gradient Methods: Convergence Results and Application to Distributed Optimization
Stochastic gradient methods for machine learning and optimization problems are usually analyzed assuming data points are sampled with replacement. In practice, however, sampling without replacement is very common, easier to implement in many cases, and often performs better. In this paper, we provide competitive convergence guarantees for without-replacement sampling, under various scenarios, f...
متن کامل